Robust Generalized Low Rank Approximations of Matrices
نویسندگان
چکیده
In recent years, the intrinsic low rank structure of some datasets has been extensively exploited to reduce dimensionality, remove noise and complete the missing entries. As a well-known technique for dimensionality reduction and data compression, Generalized Low Rank Approximations of Matrices (GLRAM) claims its superiority on computation time and compression ratio over the SVD. However, GLRAM is very sensitive to sparse large noise or outliers and its robust version does not have been explored or solved yet. To address this problem, this paper proposes a robust method for GLRAM, named Robust GLRAM (RGLRAM). We first formulate RGLRAM as an l1-norm optimization problem which minimizes the l1-norm of the approximation errors. Secondly, we apply the technique of Augmented Lagrange Multipliers (ALM) to solve this l1-norm minimization problem and derive a corresponding iterative scheme. Then the weak convergence of the proposed algorithm is discussed under mild conditions. Next, we investigate a special case of RGLRAM and extend RGLRAM to a general tensor case. Finally, the extensive experiments on synthetic data show that it is possible for RGLRAM to exactly recover both the low rank and the sparse components while it may be difficult for previous state-of-the-art algorithms. We also discuss three issues on RGLRAM: the sensitivity to initialization, the generalization ability and the relationship between the running time and the size/number of matrices. Moreover, the experimental results on images of faces with large corruptions illustrate that RGLRAM obtains the best denoising and compression performance than other methods.
منابع مشابه
Monotonicity and Iterative Approximations Involving Rectangular Matrices
A new characterization of row-monotone matrices is given and is related to the Moore-Penrose generalized inverse. The M-matrix concept is extended to rectangular matrices with full column rank. A structure theorem is provided for all matrices A with full column rank for which the generalized inverse A+ & 0. These results are then used to investigate convergent splittings of rectangular matrices...
متن کاملRobust H_∞ Controller design based on Generalized Dynamic Observer for Uncertain Singular system with Disturbance
This paper presents a robust ∞_H controller design, based on a generalized dynamic observer for uncertain singular systems in the presence of disturbance. The controller guarantees that the closed loop system be admissible. The main advantage of this method is that the uncertainty can be found in the system, the input and the output matrices. Also the generalized dynamic observer is used to est...
متن کاملLow-Rank Matrices on Graphs: Generalized Recovery & Applications
Many real world datasets subsume a linear or non-linear low-rank structure in a very low-dimensional space. Unfortunately, one often has very little or no information about the geometry of the space, resulting in a highly under-determined recovery problem. Under certain circumstances, state-of-the-art algorithms provide an exact recovery for linear low-rank structures but at the expense of high...
متن کاملLow Rank Approximation using Error Correcting Coding Matrices
Low-rank matrix approximation is an integral component of tools such as principal component analysis (PCA), as well as is an important instrument used in applications like web search, text mining and computer vision, e.g., face recognition. Recently, randomized algorithms were proposed to effectively construct low rank approximations of large matrices. In this paper, we show how matrices from e...
متن کاملInexact and incremental bilinear Lanczos components algorithms for high dimensionality reduction and image reconstruction
Recently, matrix-based methods have gained wide attentions in pattern recognition and machine learning communities. The generalized low rank approximations of matrices (GLRAM) and the bilinear Lanczos components (BLC) algorithm are two popular algorithms that treat data as the native twodimensional matrix patterns. However, these two algorithms often require heavy computation time and memory sp...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره 10 شماره
صفحات -
تاریخ انتشار 2015